Online Learning from Finite Training Sets: An Analytical Case Study
نویسندگان
چکیده
We analyse online learning from finite training sets at noninfinitesimal learning rates TJ. By an extension of statistical mechanics methods, we obtain exact results for the time-dependent generalization error of a linear network with a large number of weights N. We find, for example, that for small training sets of size p ~ N, larger learning rates can be used without compromising asymptotic generalization performance or convergence speed. Encouragingly, for optimal settings of TJ (and, less importantly, weight decay ,\) at given final learning time, the generalization performance of online learning is essentially as good as that of offline learning.
منابع مشابه
On-line Learning from Finite Training Sets in Nonlinear Networks
Online learning is one of the most common forms of neural network training. We present an analysis of online learning from finite training sets for non-linear networks (namely, soft-committee machines), advancing the theory to more realistic learning scenarios. Dynamical equations are derived for an appropriate set of order parameters; these are exact in the limiting case of either linear netwo...
متن کاملOnline Learning from Finite Training Sets and Robustness to Input Bias
We analyze online gradient descent learning from finite training sets at noninfinitesimal learning rates eta. Exact results are obtained for the time-dependent generalization error of a simple model system: a linear network with a large number of weights N, trained on p = alphaN examples. This allows us to study in detail the effects of finite training set size alpha on, for example, the optima...
متن کاملOnline learning from nite training sets
We analyse online gradient descent learning from nite training sets at non-innnitesimal learning rates for both linear and non-linear networks. In the linear case, exact results are obtained for the time-dependent generalization error of networks with a large number of weights N, trained on p = N examples. This allows us to study in detail the eeects of nite training set size on, for example, t...
متن کاملDynamical and stationary properties of on-line learning from finite training sets.
The dynamical and stationary properties of on-line learning from finite training sets are analyzed by using the cavity method. For large input dimensions, we derive equations for the macroscopic parameters, namely, the student-teacher correlation, the student-student autocorrelation and the learning force fluctuation. This enables us to provide analytical solutions to Adaline learning as a benc...
متن کاملSupervised learning with restricted training sets: a generating functional analysis
We study the dynamics of supervised on-line learning of realizable tasks in feed-forward neural networks. We focus on the regime where the number of examples used for training is proportional to the number of input channels N. Using generating functional techniques from spin glass theory, we are able to average over the composition of the training set and transform the problem for N → ∞ to an e...
متن کامل